List of AI News about prompt caching
| Time | Details |
|---|---|
|
2026-02-21 18:06 |
OpenClaw v2026.2.21 Release: Gemini 3.1 Integration, Security Hardening, and Discord Voice Streaming — Detailed AI Platform Analysis
According to @openclaw, the v2026.2.21 release delivers Gemini 3.1 integration, major security hardening, Discord streaming with voice channels, thread-bound subagent sessions, iOS and Apple Watch UX polish with gateway stability, and prompt caching tweaks, alongside 100+ fixes (as linked in the OpenClaw GitHub release notes). As reported by the OpenClaw GitHub repository, the Gemini 3.1 addition positions OpenClaw as a multi-model orchestration layer for production agents, while the security overhaul and session isolation directly target enterprise adoption and compliance needs. According to the release notes, Discord voice streaming plus thread-scoped subagents expands real-time agent collaboration and moderation workflows, and prompt cache improvements reduce latency and inference costs—key business levers for customer support, community management, and on-device assistant scenarios. |
|
2025-11-14 00:55 |
OpenAI Launches Advanced API Models and Extended Prompt Caching for Developers: Boosting AI Application Performance in 2025
According to Greg Brockman on X (formerly Twitter), OpenAI is prioritizing developer needs by introducing excellent new models in their API and implementing extended prompt caching features (source: x.com/gdb/status/1989135114744573993). These advancements are designed to accelerate AI-powered application development and improve efficiency for businesses leveraging OpenAI's platform. By providing more robust model options and optimizing prompt caching, OpenAI enables developers to build faster, more reliable, and cost-effective AI solutions. This move positions OpenAI to better serve enterprise and startup clients seeking scalable, production-ready generative AI tools (source: x.com/gdb/status/1989135114744573993). |
|
2025-11-13 19:11 |
GPT-5.1 API Release: New Features, Codex Models, and 24-Hour Prompt Caching for Developers
According to Sam Altman (@sama), OpenAI has launched GPT-5.1 in the API, maintaining the same pricing as GPT-5. The release includes specialized models gpt-5.1-codex and gpt-5.1-codex-mini, designed for long-running coding tasks, which expands AI-driven software development capabilities. Additionally, prompt caching duration has been extended to 24 hours, significantly optimizing operational efficiency for businesses managing large-scale AI workflows. This update offers practical advantages for enterprise AI adoption, enabling more reliable automation and cost-effective coding solutions (Source: @sama, Nov 13, 2025). |
